Cautious weighted random forests

نویسندگان

چکیده

Random forest is an efficient and accurate classification model, which makes decisions by aggregating a set of trees, either voting or averaging class posterior probability estimates. However, tree outputs may be unreliable in presence scarce data. The imprecise Dirichlet model (IDM) provides workaround, replacing point estimates with interval-valued ones. This paper investigates new aggregation method based on the theory belief functions to combine such intervals, resulting cautious random classifier. In particular, we propose strategy for computing weights minimization convex cost function, takes both determinacy accuracy into account it possible adjust level cautiousness model. proposed evaluated 25 UCI datasets demonstrated more adaptive noise training data achieve better compromise between informativeness cautiousness. • A classifier forests proposed. Tree are automatically learned from using function. tuned single parameter. Extensive experiments demonstrate interests method.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

News Articles Classification Using Random Forests and Weighted Multimodal Features

This research investigates the problem of news articles classification. The classification is performed using N-gram textual features extracted from text and visual features generated from one representative image. The application domain is news articles written in English that belong to four categories: Business-Finance, Lifestyle-Leisure, Science-Technology and Sports downloaded from three we...

متن کامل

Hybrid weighted random forests for classifying very high-dimensional data

Random forests are a popular classification method based on an ensemble of a single type of decision trees from subspaces of data. In the literature, there are many different types of decision tree algorithms, including C4.5, CART, and CHAID. Each type of decision tree algorithm may capture different information and structure. This paper proposes a hybrid weighted random forest algorithm, simul...

متن کامل

1 Random Forests - - Random Features

Random forests are a combination of tree predictors such that each tree depends on the values of a random vector sampled independently and with the same distribution for all trees in the forest. The generalization error for forests converges a.s. to a limit as the number of trees in the forest becomes large. The error of a forest of tree classifiers depends on the strength of the individual tre...

متن کامل

Mondrian Forests: Efficient Online Random Forests

Ensembles of randomized decision trees, usually referred to as random forests, are widely used for classification and regression tasks in machine learning and statistics. Random forests achieve competitive predictive performance and are computationally efficient to train and test, making them excellent candidates for real-world prediction tasks. The most popular random forest variants (such as ...

متن کامل

Random Composite Forests

We introduce a broad family of decision trees, Composite Trees, whose leaf classifiers are selected out of a hypothesis set composed of p subfamilies with different complexities. We prove new data-dependent learning guarantees for this family in the multi-class setting. These learning bounds provide a quantitative guidance for the choice of the hypotheses at each leaf. Remarkably, they depend o...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Expert Systems With Applications

سال: 2023

ISSN: ['1873-6793', '0957-4174']

DOI: https://doi.org/10.1016/j.eswa.2022.118883